Speech to Text Model, where the model doesn't attempt to correct errors/grammer? - artificial-intelligence

Is there an Vosk speech to text model, or any other open source/closed sourced model, where the model would output the spoken words into text. But it wouldn't correct them into proper words or fix their grammer, just output what they are saying in english/gibberish?
Currently, Vosk English level models and Deepspeech models output and accurately correct what the user is saying, but I am trying to find the opposite where it just outputs what is spoken.
If there are no models already trained for this, how would I go to getting one to behave like this?

Related

How can I make a photograph database that displays the results of a query as some kind of image gallery?

I am taking a database design class and for a project want to make a database of my mom's digital photos for her. I haven't dealt in application up to this point, only theory, but I have Access. Therefore, ideal answers don't suggest non-database solutions and don't assume I know much about actual database implementation. Solutions specific to Access could also be a plus. I hope that precursor saves some time and effort.
Theoretically, my mom wants to see all photos of pets from '05-'07 in raw format, and she enters an appropriate query. I suspect I can handle it up to there. However, at the moment, the best I can figure out to do is to return a column of either attachments or OLE objects. 5 clicks per photo is not ideal. I need a faster way to present the images. Opening them all in a grid of thumbnails or as a one-click-slide-show would seem the natural fit, but whatever works. How can I accomplish this?
Less important but worth consideration is the fact that, at some point, it would be great if this same type of system could be implemented on the internet for all of the family reunion photos she has taken, but I will take what I can get.
Use one form to get parameters for the query. then use another form(more processing) or report(if printing) to show the selected pictures. I will not cover passing parameters but here are some links.
https://www.fmsinc.com/microsoftaccess/forms/openargs/index.htm
https://learn.microsoft.com/en-us/office/vba/access/concepts/forms-design/apply-a-filter-when-opening-a-form-or-report
There is a complication, in Access pictures are usually stored in the attachment type. the attachment column can hold many pictures in each record. So if we have a table called Pictures with an attachment type column also called Pictures, then each individual picture is actually stored under Pictures.Pictures.FileData.
So to display the picture query we use a form/report with default view set to Continuous Forms (displays many records or in this case pictures on the same page) then in the details section of our Display form we place an attachment control and bind that control to our filtered Pictures.FileData.
Format and add functionality to taste.

Supervised Learning with Form Recognizer

I am trying to analyze a form using Microsoft's Form Recognizer API but I am not seeing the results I had hoped for. After training the model on my form, the keys it has generated are very rarely what I want to be. Does anyone know if there is a method to improve accuracy of key recognition? I was thinking there might be some way to give a list of key/value pairs when training as a form of supervised learning.
Here is a sample of the form I'm trying to parse.
I'd expect keys of 'Year', 'Make', 'Model', and 'VIN'. But instead the model is returning a key of 'Vehicle' with values 'Year', 'Make', 'Model', and 'VIN' and their subsequent values.
I know I specifically asked about supervised learning but really any techniques or tips on how to improve the accuracy of a form recognizer model would be appreciated.
Azure Form Recognizer now offers a Supervised Learning Tool to tune models for forms which are difficult to train with the default unsupervised learning mode.
Here's how you can get to the tool:
https://learn.microsoft.com/en-us/azure/cognitive-services/form-recognizer/quickstarts/label-tool
Did you train a model with 5 sample forms ? Can you try adding an empty form without the values to the training data and see if it helps ? Are the forms good quality scanned or are they tilted ?
Following are some tips on how you can improve the accuracy:
How to Build a training data set for a custom model
When you use the Form Recognizer custom model, you provide your own training data so the model can train to your industry-specific forms. You can train a model with five filled-in forms or an empty form (include the word "empty" in the file name) plus two filled-in forms. Even if you have enough filled-in forms to train with, adding an empty form to your training data set can improve the accuracy of the model.
It's important to use a data set that's optimized for training. Use the following tips to ensure you get the best results from the Train Model operation:
• If possible, use text-based PDF documents instead of image-based documents. Scanned PDFs are handled as images.
• Use one empty form and two filled-in forms if you have them available.
• For filled-in forms, use examples that have all of their fields filled in.
• Use forms with different values in each field.
• If your form images are of lower quality, use a larger data set (10-15 images, for example).

Parse - how to add comments to an image?

I'm trying to create an iOS application where users post and image, and then someone can comment on the image. So far I've set up the a message class which contains (among other things) the image file. I want to be able to add comments to the image, so what would be the best way of doings this.
Option A:
I could add an array (or object?) column to the already existing message class, and then store in this array the comment (string) and the id of the poster (string). For this I think I would need a two-dimensional array, but I'm not sure how I would go about doing this.
Option B:
I could make an entire new class of comments which contains the user's comments (string), as well as the image file that he/she linked to (perhaps though a PFRelation)
Basically I'm leaning to Option A, because it seems easier/more efficient to implement, but I don't really know how I would go about creating a two-dimensional/array of objects - so my question is, how would I go about doing this?
The way I did it in my app is option B. I then added a count column to the photo that was incremented/decremented onSave or onDelete. I think this method is better because it keeps the data more symmetric (with option A some data could have thousands of comments while others would have 0). Additionally, it allows you to query the comments cell if you wanted to have a notification view where you showed the user what activity has been done on their photo.

CRM 2011 Search Exact Match

I have a problem with CRM 2011 , I currently have 22 users . In the Accounts or Cases entities some of them can search with half of the word and some of them have to enter the whole exact name to get answers which is really annoying . Can someone help please?
Thank You !
By default you can add which fields will be searched to the Quick Search view, see how Example.
You still can use wild cards "*" (ex: "John") when performing a Quick Find Search.
Make sure they use all the same view (the Quick find one usually) and, from
Settings->Customizations->Customize the system->Entities->(Select the desired entity)->Views and then open the view users are using then you can set the filter criteria to whatever you would like for each fields.
I don't think you can enforce a view on users though, but my guess if for some users partial searches work, and for some others not, they are probably using different views.
EDIT: You can select said view and in More Actions above set it to default view. Then again, users are still able to create personal views and set them as default...
Surely you mean that your users search their wanted fields in quick find area that this is system view and the problem posed do not have relation to view's setting, because it was the same for all the user's.
so i guess this problem arises out: keyboard not uniform.......

CakePHP Parsing data from file and adding to database

I have a file that I wish to parse and add to the database.
The way I wish to do it is this:
The user uploads the file
File is parsed
The users are given the parsed data and selects what they want to keep
They submit
(More detailed explanation follows)
I have a film script file that I wish to parse and enter scenes and their respective characters and props. I have already managed to parse this file into an array using php.
What I want to do now is to
prompt the user with the various scenes and its contents
allow the user to fix / modify the data before submitting it into the database.
The thing is that the script file has more than one scene in it, so I am unsure how to handle the data-entry aspect of this problem.
Are there any "best practices" for these kinds of data entries?
(if I didn't explain something in enough detail, please tell me and I'll update the question).
I would simply parse the file and save it as it is to the database. Something like Script hasMany Scenes. Let them select their script and simply paginate the Scene records and let them delete them.
Is there an urgent need to not save them all after import? Another downside of your approach is that if the browser crashes or the user closes it he has to start over again because you just kept the data in memory but not persistant.
If you want to rely on your approach you'll have to work with the data coming back from the form and simply remove the key from the form data.
Like doing this in a foreach running over the data structure
if ($data['Scene][$key]['delete'] == 1) { unset($data['Scene'][$key]); }

Resources