None of that made Tesseract excel in capturing handwritten text…
None of that made Tesseract excel in capturing handwritten text…
It wasn’t possible a year ago when pos6ted around with tesseract. Things might have changed during the last couple of months though.
To train an AI to recognize handwriting you need a huge dataset of handwriting examples. That is millions of samples of handwritten text + information about what the written text says in every example).
This is why the best engines only exists as a service in the cloud. The OCR engines you can install lovely that are acceptable, but far from perfect, are commercial. Parascript FormXtra is one of the better commercial ones.
The only OCR Engine that’s free and really good is Tesseract OCR but it doesn’t handle handwritten text.
Not even if Windows and Linux were on different partitions on the same disk would Windows be able to access the files on the Linux partition without the key.
Just pointing out that s separate disks doesn’t change anything. The data, in its encrypted form, will be inaccessible without the decryption key.
Yup. /r/Datahoarder guided me right. Got two of the recommended model of MyBook and shucked them. This was 2-3 years ago. Disks are still going strong in my NAS.
You seeing me as aggressive just because I’m right about Firefox being a web browser is nothing I can do anything about and something you have to work with.
Me and Mozilla will keep trucking and call Mozilla Firefox a Web Browser… Or as Wikipedia says:
Mozilla Firefox, or simply Firefox, is a free and open-source web browser developed by the Mozilla Foundation and its subsidiary, the Mozilla Corporation.
And you keep calling Firefox your banana or whatever. It’s ok. I promise to continue to not being aggressive about it.
Banana or not, a document viewer / editor that handle a subset of two standards is not a very versatile document viewer / editor.
So back to my original point: Since Mozilla doesn’t have a very stable business model it seems dangerous to focus on other things than making their web browser the best at browsing the web.
Ps. It took us over 20 years to get rid of the cluster f***s Internet Explorer and Flash was and it seems we should have learned the lesson by now. Going down the same route as before, starting to support standards that rely on patents owned by a third party (Adobe in this case) is definitively not a death sentence in any way, but history has shown us that it’s a slippery slope that has many different paths and endings.
You obviously know very little about PDF since you’re trying to put it on the same level as HTML.
The PDF files that we see in our daily life today is not even halfway as open as HTML is and dealing with them is not as easy as you think.
Furthermore, you’re free to call Firefox a “document viewer”.
I personally prefer a state of the art web browser and a state of the art PDF viewer more than one document viewer that only handles subsets of 2 document types.
That was 20 years ago. Numerous PDF readers has surfaced since then.
PDF is not like HTML.PDF is a messy standard where you need Adobe products to support all the shit that a PDF could contain. There is no open source product that for example fully support PDF forms and therefore Mozilla won’t either.
All in saying is that Firefox is a web browser and not a document viewer. Since Mozilla would go bankrupt in two hours if Google stopped showing them with a shitload of money, Mozilla would be wise to focus on the core.
So if Google were to stop paying Mozilla for us to be able to use Firefox for free, we’re all running Chrome.
Let’s hope that if they support the shit show XFA is, they manage to support it fully.
The only one that seem to support it fully today is Adobe and I still haven’t been able to find any open source product that was able to close forms in PDFs.
I just wish Mozilla got back to focus on making the best browser.
PDF stuff is cool but not why I use a web browser.
Reported, rule 5
That chart doesn’t say anything about system resource usage.
Edit: found the performance chart now. Still no explanation on what performance tests(more than two sentences) they performed and how the scoring was applied.
Interesting. Do you have links that support your claims that I can read up on?
Yeah, that was you continuing to show how inexperienced you are.
For a remote exploit to work the computer or device has to expose ports to the network your computer is connected to.
“Remote” means that the vulnerability does not require local access. So if your friend connects his infected device to your wifi, all devices connected to the same network essentially are at risk, depending on what’s listening on the devices and what vulnerabilities they have.
Your idea about avoiding bad websites is ridiculous. History is full of examples where third party ads had been created to infect one way or another. That’s ads that users on legitimate site were exposed to. That’s just one little example. There have been numerous examples of malicious sleeping JavaScript code that suddenly wakes up and contacts it’s command-and-control server and then download malicious JavaScript code to unknowing site visitors.
Furthermore, you didn’t understand my question. Of course antivirus is able to stop malware it recognizes that enters through a remote exploit. The user with antivirus would at least have a chance of knowing that something was up each time and attempt to infect was made.
You on the other hand would sit there clueless with your little zombie computer and laugh at all them script kiddies.
But hey… You just continue trying to infect others around you with bad security advice and have a good day. I’m outta here.
I explained what a remote exploit was and gave examples of remote exploits.
Are you claiming that antivirus isn’t able to detect malware entering through an remote exploit?
Either you’re just ignorant or your working in the Russian malware industry.
Remote exploits doesn’t have anything to do with you running any infected executables. It’s about vulnerabilities in executables that you are running. Read up on the zx vulnerability or the log4j vulnerability.
One really really old attack vector is a buffer overflow attack. For example, if you’re running a clean VLC to watch a movie and your VLC is older than version 3.0.12 you’re at risk. The video file, that you “purchased” on PirateBay, could have been manipulated to crash VLC and force VLC run a specific payload in the video file. If that payload is ransomware it’s game over for you.
Yeah, just like wearing a seatbelt doesn’t guarantee that you don’t get injured, antivirus doesn’t guarantee that your computer won’t get infected.
But there’s no doubt about the usefulness of both seatbelt and antivirus.
I fully agree.
But my main point was that they’re taking an extreme risk if they’re running without active antivirus and access the network in one or another way.
Yes. But the discussion was about not running any since it killed performance.
If your computer is network connected you really really need antimalware running. In theory, a game server could be exploited and controlled to inject malware into game clients ( = you playing online).
If you use a browser to access internet, there has been malware infected ads that infect your computer when you visit legitimate web sites.
If you get infected, the malware most likely won’t do anything that makes you notice it. It all depends what the purpose of the malware is.
Ransomware? Then it stays hidden until it has encrypted all your files and then it pops up telling you to pay or you won’t be able to use your computer.
Collecting useful passwords? Getting full access to your Steam account is nice. Got some awesome weapons and armor in the MMORPG you play? That’s something they theoretically would script to be able to steal from you.
Or maybe the malware just stays hidden for now, contacting it’s control center now and then to see if there is any instructions.
Malware is business. The people behind it are businessmen and you are part of what they sell.
If you have a fairly modern computer there shouldn’t be that much impact on the performance.
How good is good do you say?
We got a pretty good results with CER at 4% and WER at 15%!
This was on a limited dataset used to test and train which most likely means that if you introduced an even larger dataset with greater variations in handwriting style for testing the numbers might be even worse.
Very simplified: A risk of a character wrong every 20th character and a word wrong every 7th word. The SER was around 20%.
There’s an reason why no one has released a good model for western letters yet and why companies pay up to 1€ for capturing data from 10 handwritten pages.
It will come but OCR isn’t as sexy as developing text2image solutions.