I'm trying to export the data from Webalizer's reports, but i have a problem when it comes to the data of each country.
Webalizer generates an image containing a graph of countries statistics, but I have not been able to find any file where i can find this data besides that image.
Does Webalizer saves this data anywhere?
Yes, webalizer data can be found here: /home/youruser/domains/example.com/stats/ if you're using a centos server with directadmin, Browse your folders to find the location on your system as it can be somewhere else on your machine depending on your configuration.
Related
Is there really no way to export/import the adaptive junk filter training between computers?
I just setup thunderbird on a new computer and I feel like the junk filtering is sorely lacking behind my other workstation where the adaptive junk filter has had quite some time to train. I can't find anything on it in the options or while googling.
Thanks in advance
Some links:
How the junk filtering works (bayesian filtering)
Junk Mail Controls feature description (old as mold)
You should be able to transfer this data by copying the file training.dat from your old profile to your new one.
Be sure to close both instances and backup your Profile folders before copying files from one profile to another.
You can also see Mozilla Support on how to find your Profile folder and Mozilla Zine for a full list of files in your Profile folder.
I have challenges in fetching picture saved as bytes in sqlite database in codename one. Pls is there any special way in codename one to get picture from the sqlite database?
I am using this code
byte[] img = r.getBlob(0);
pls is their anythng wrong here?. I need a snippet code that will fetch pictures from the database. Bcos my pictures must reside inside database for better calling. Thanks
I would avoid that and instead store the data as a file as the blob functionality is flaky across platforms. Save the path to the file (or its name) in sqlite and store the files in storage or filesystem.
I am a complete newbie when it comes to web page design, and what I am trying to achieve is a web page that I can display on a wall mounted screen as an office dashboard. I have data in excel that is constantly being updated (on the server) and I want to be able summarise this and display it (e.g. total orders etc.) for staff to see. Therefore the web page needs to be able to connect to the data source, and update itself every few minutes. I am hoping to then use Ubuntu or even Raspberry Pi to drive the dashboards.
Can anybody point me towards either some clear instruction on how to achieve this, or better still some sample files that will help me see how its done?
Really appreciate any help!!
If you want to use PHP, you can use PHPExcel, to read your Excel files, and if you want to only display the information as is, you may output it to HTML directly, there is an API to manipulate the Excel file, so you may use that to summarize your data, however, if you need something cleaner, you may want to use Windows instead of Linux, given that in Windows, you can use an Excel file as a data source, and there are third party products that can use them as if it were a database (using SQL queries to retrieve data).
I am helping out an organization which are planning on changing their members system. Right now their system is developed in Plone and all their data is in a Data.fs file.
Their system is down for the moment and it would take some time and effort to get it up and running.
Is there a way to get the data out from the database into a standard format such as csv files or SQL? Or do they need to get the system up and running beforehand and export the files from "within" plone?
Thanks for your help and ideas!
Kind regards,
Samuel
The Data.fs file is a Object Oriented Database file, and it is written by a framework called the ZODB. The data within it represent python instances, layed out in a tree structure.
You could open this database from a python script, but in order for you to make sense of the contained structures, you'll need access to the original class definitions that make up the stored instances. Without those class definitions all you'll get is placeholder objects (Broken objects) that are of no use at all.
As such, it's probably easier to just get the Plone instance back up and running, as it'll be easier to export the exact data you want out if you have things like the catalog (basically a specialized database index) to build your export.
It could be that this site is down because of something trivial, something we can help you with here on Stack Overflow, or on the Plone users mailinglists or in the #plone IRC channel. If you do get it up and running and have some details on what you are trying to export, we certainly can help.
You'll need to get the system up and running to export data. Data in the data.fs file is stored as Python pickles and is not intelligible to "outside" systems.
As the others have pointed out before, your best course would be to have Plone running back again. After doing so, try csvreplicata to export existing data to csv format. And for user accounts, try atreal.usersinout.
If you need professional help, you can search for available providers from http://plone.org/support/providers
For free support, post specific problems here.
Recently I managed to export Plone 4 site to sqlite using SQLExporter: http://plone.org/products/proteon.sqlexporter. But you need to get your Plone instance working first to use it.
I am interested in creating a video databse. My goal is to have a folder where my videos will be kept and each time I copy/delete a video the website that presents them should be updated to. the problem is I have no idea how to approach it.
Should I..
Use Sql and store a reference to each video location?
Have a script that checks all the time if new changes happen in that folder?
A package like joomla?
I am using ubuntu btw. I already have a simple html5 page, and I am presenting the videos using html5 video.
It depends on the size and the performance you want.
1.Way : use php to scan the folder and generate links on the fly
2.way : Use a database to store the file names and retrieve the names from the database and generate urls
pros and cons.
simple to implement , no changes in upload or download script. no database required.
You need have a database , little coding required for upload and also while genrating a page
You should make a db (format does not matter) and storing in it only file names of videos: the videos would be stored on hard drive.
Any operation on the web site will pass first on db for insert/update/delete videos records and then (maybe in a transaction context) on the file system.
This would be the standard approach to your question.