
Mailing List Archive
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [tlug] Anyone alive out here ?
On 2024/09/05 20:58, Darren Cook wrote:
Later, I installed ollama and blogged about my experience with a few
freely available LLMs:
https://mstdn.io/@codewiz/112527717194517544
That was interesting. Which AMD GPU are you using?
Radeon RX 7900 XT. It's 2 years old, but very stable and performs well
on Linux for desktop, gaming and compute workloads. AMD's
CUDA-compatible stack no longer sucks as it used to.
The main issue is having only 20GB of RAM. Not sure when they will
announce the next gen of desktop GPUs with at least 40GB, which would be
very useful for LLM inference and training.
If you need more GPU RAM, I'd recommend getting a VM on Google Cloud or
other provider that charges by the hour.
What's the current state of the art? Interfacing models with text has
big limits...
Quite closely related, I've been wondering what the state of the art for
open-source OCR is, particularly of Japanese text.
I'm waiting for the first Llama-like LLM with image recognition similar
to ChatGPT.
It's not the same of classic OCR and has obvious limitations, but a
multimodal LLM can figure out things from context, like "oh, this is the
hand writing of a young child... and this word is misspelled, I think
they meant to write hamburger... pages 3 and 4 appear to be swapped".
> All the links go to Tesseract (or wrappers around it), which is simply
> not good enough, even for English. Or to online tools where you upload
> your private data, and pay for the privilege.
>
> This could then lead on to the greatest unsolved computing challenge
> of the 21st century, which is a PDF to text converter. Yes, yes, I
> know toy examples work. I mean PDFs that were made in Microsoft Word,
> contain multiple columns and figures, and inset boxes, like a
> magazine, or a company's annual report. And extracting all the
> sections with headings, in a reasonable reading order (suitable for,
> say, a screen reader).
>
> I thought Google, OpenAI, etc. had solved it, as they train the LLMs
> on PDFs, and ChatGPT can be given a PDF and answer questions. But, as
> far as I've been able to find out, they either feed raw PDF bytes in,
> and hope, or they use pdf2text, and hope.
Ultimately, narrow models for OCR, speech recognition and translation
can't improve beyond a certain point because they lack semantic
understanding. They also can't be instructed in plain English to adjust
the output (e.g. "use informal language in the dialogue part, but formal
Japanese for the professor").
Just yesterday, I asked ChatGPT to compare a 5-page lease contract in
Japanese with an older one from 5 years ago, then summarize the
important differences along with any clauses I should watch out for.
I could have asked "translate this PDF to English", but then I would
have to do all the hard work by myself :-)
--
_ // Bernie Innocenti
\X/ https://codewiz.org/
Home |
Main Index |
Thread Index