314 points by lermontov 2 days ago | 123 comments
mmastrac 2 days ago
https://github.com/mmastrac/gibbet-hill
Top and bottom halves of the page in the repo here:
https://github.com/mmastrac/gibbet-hill/blob/main/scan-1.png https://github.com/mmastrac/gibbet-hill/blob/main/scan-2.png
EDIT: If you have access to a multi-modal LLM, the rough transcription + the column scan and the instruction to "OCR this text, keep linebreaks" gives a _very good_ result.
EDIT 2: Rough draft, needs some proofreading and corrections:
quuxplusone 2 days ago
(I've transcribed various things over the years, including Sonia Greene's Alcestis [1] and Holtzman & Kershenblatt's "Castlequest" source code [2], so I know it doesn't take much except quick fingers and sufficient motivation. :))
[1] https://quuxplusone.github.io/blog/2022/10/22/alcestis/
[2] https://quuxplusone.github.io/blog/2021/03/09/castlequest/
EDIT: ...and as I was writing that, you seem to have finished your transcription. :)
mmastrac 2 days ago
I've done transcription in the past myself (did two books for standard ebooks with some from-scratch transcription and lots of editing) and I know the pain. I've always found it easier to fix up OCR than type the whole thing by hand because I've found my error rate of eyeball transcription to be higher.
If you want to tackle the proofing passes, I'm happy to add you to the repo :)
wahnfrieden 1 day ago
CoastalCoder 20 hours ago
I know nothing about OCR, so maybe it's obvious to others.
dylan604 16 hours ago
wahnfrieden 15 hours ago
dylan604 13 hours ago
also, if it is full local and offline why does it need to be rented?
wahnfrieden 12 hours ago
wahnfrieden 15 hours ago
eru 1 day ago
LLMs are increasingly becoming cheaper and more accessible than humans with a baseline of literacy.
notachatbot123 22 hours ago
*: If you ignore the environmental costs.
eru 21 hours ago
They are better than me at many tasks.
> Not everything has to be solved by cheap* technological processes.
> *: If you ignore the environmental costs.
For many tasks, inference on an LLM is a lot cheaper (including for the environment) than keeping a human around to do them. As a baseline, humans by themselves take around 100W (just in food calories), but anyone but the poorest human also wants to consume eg housing and entertainment that consumes a lot more power than that.
homebrewer 20 hours ago
eru 6 hours ago
We can reduce the environmental footprint of specific activities by replacing humans. Yes, we would only reduce the _overall_ footprint by reducing the number of humans.
CoastalCoder 20 hours ago
I'm assuming that powering them down isn't a viable option, unlike with GPUs in a datacenter.
eru 1 hour ago
Presumably the humans would be enjoying with some other activity. Eg they could be working on carbon capturing projects? Or producing electric power via pedaling, etc. I don't know.
I was purely talking about the environmental impact of this one activity.
throwaway0123_5 18 hours ago
Sadly that might be assuming too much... here and on reddit I've seen a handful of people who have said that we should continue with AI progress even if it causes the extinction of humans, because we'll have ~"contributed to spreading intelligence throughout the universe and it doesn't really matter if it is human or not."
With that as the extreme end of the spectrum, I suspect the group of people who simply aren't considering what happens to obsoleted humans is much larger, and corporations certainly haven't demonstrated much interest in caring for those who technology has obsoleted in the past.
Tbh it is really disheartening to see so many technologists who seemingly only care about technology for its own sake.
arp242 13 hours ago
Obviously not true because that human is alive regardless, and has mostly the same base energy needs no matter what they're doing.
Reducing humans to just energy-using machines is an absolutely insane misanthropic take.
eru 6 hours ago
The human would presumably do something else they enjoy doing more, if a machine took that specific job.
cxr 1 day ago
<https://woodsfae.tumblr.com/post/764918993659330560/gibbet-h...>
oliyoung 1 day ago
That's like bringing a knife to a gun fight my friend, never underestimate the power of a committed Tumblr user
CoastalCoder 20 hours ago
(Southpark homage)
drivers99 1 day ago
simonw 2 days ago
mmastrac 2 days ago
https://github.com/mmastrac/gibbet-hill/blob/main/col-1-a.pn...
reaperducer 2 days ago
Tip: Load the pngs into Preview, hit "Auto Levels," and crank up "Sharpness" on each one. Looks pretty good!
qmr 2 days ago
staticman2 2 days ago
red369 1 day ago
I realise I've used vague terms in that sentence, even setting aside the tricky question of what makes the things often described as great works "greater" than things that are looked down on, but might be much more popular.
I once read a great foreword to a novel lamenting the loss of "good bad books", citing Dracula as an example. It was by a famous author (as I remember), but I can't remember, and can't find, the foreword or the novel I'm thinking of.
inejge 16 hours ago
It's an article/essay by George Orwell (actually titled Good Bad Books), available online.
latexr 19 hours ago
It’s interesting that Dracula falls right in the middle of his career. It strongly suggests it was a fluke. Doesn’t look like he ever had an inkling of how famous the story would be, the Wikipedia page says he was best known while living for being a personal assistant and business manager to some other bloke. That’s a bit sad.
pozdnyshev 18 hours ago
nu11ptr 1 day ago
red369 1 day ago
By the way, for anyone who is thinking of reading Dracula's Guest, it is likely it was intended as a first chapter of Dracula, but was cut.
xenospn 20 hours ago
paulette449 23 hours ago
barrkel 20 hours ago
politelemon 2 days ago
Go full screen and go to page 2 it starts at about the middle.