Hacker Newsnew | past | comments | ask | show | jobs | submit | iib's commentslogin

This seems really nice, and looks like something I have been wanting to exist for some time. I will definitely play with it when I have some time.

I know this is a personal project and you maybe didn't want to make it public, but I think the README.md would be better suited with a section about the actual product. I clicked on it wanting to learn more, but with no time to test it for now.


Thanks for the feedback, I did update the README and included all the futures and also there is https://talimio.com, I think it shows the future in a better way visually


Didn't see the website at first. Thank you!


That sounds awesome. But I have two curiosities: What are the problems of httpx? And was pycurl not enough for what you wanted to do?


Httpx/httpcore are abandoned and their connection pooling etc implementations are completely broken https://github.com/encode/httpx/issues/3215 Many PRs trying to fix varying issues https://github.com/encode/httpcore/pulls But maintainer(s) are not intrested in fixing any of the issues.


pycurl doesnt support async, right?


I have been looking for the same thing, either from Meta's SAM 3[1] model, either from things like the OP.

There has been some research specifically in this area with what appears to be classic ML models [2], but it's unclear to me if it can generalize to dances it has not been trained on.

[1] https://ai.meta.com/blog/segment-anything-model-3/

[2] https://arxiv.org/html/2405.19727v1


For small edits, has anybody configured a leader-key scheme? Something like Doom Emacs has with space as a leader.

It seems to me to be the best possible configuration for Emacs on Android (on a phone) and I was wondering if I should invest time in such a solution.

strokes-mode.el would also be very nice, but apparently it doesn't have touchscreen support.


As soon as I found out that this model launched, I tried giving it a problem that I have been trying to code in Lean4 (showing that quicksort preserves multiplicity). All the other frontier models I tried failed.

I used the pro version and it started out well (as they all did), but it couldn't prove it. The interesting part is that it typoed the name of a tactic, spelling it "abjel" instead of "abel", even though it correctly named the concept. I didn't expect the model to make this kind of error, because they all seems so good at programming lately, and none of the other models did, although they did some other naming errors.

I am sure I can get it to solve the problem with good context engineering, but it's interesting to see how they struggle with lesser represented programming languages by themselves.


Made me think of this SMBC comic[1], where there's a debate if being in English or Spanish, each with around a billion speakers, makes it rare or not.

[1] https://www.smbc-comics.com/comic/phonemes


It depends on whether you define "rare" in terms of language variety or human variety, obviously. In terms of languages, it is a relatively rare phoneme. It occurs more often as an allophone of other phonemes, but in that case the speakers may not be able to distinguish it and will struggle to reproduce it in "unusual" environments.


How do you solve the second mission (the man that is hiding somewhere)? I've managed to find even the three.js mention, but nothing to start that one.


He's in a cave near the forest. Turn right when you see the tree stump with the owl on it.


I think those kind of glasses may be really useful for blind people. I have seen similar glasses targeted at blind people, that at least in theory, seemed to me like a good idea.

I recall the glasses also can write on the screen inside the lens, which makes me think they may be good for deaf people as well.

It's just that these use-cases seem uncool, and big companies seem to have to be cool in order to keep either their status or their profits. But I have a feeling the technology may be really useful for some really vulnerable people.


Yes, there are people working on image recognition glasses for blind people.

Nobody seems to have been successful yet, and I think the focus on applying LLMs instead of dumb UI and mixed dumb and ML image processing is a large reason why.


Oh I do still enjoy the glasses, they are actually rather incredible, even though they do not have a screen. That said - These actually do have a Be My Eyes integration - It is incredibly impressive.


This is exactly the fine-tuning I am hoping for, or I would do if I had the skills. I tried it with gemma3 270M and vanilla it fails spectacularly.

Basically it would be the quickadd[1] event from google calendar, but calendar agnostic.

[1] https://developers.google.com/workspace/calendar/api/v3/refe...


How do you use the recordings as part of the CI?


We made a tool using termsvg: https://github.com/DeepGuide-Ai/dg. It'll use recorded sessions and execute the non-interactive sessions during CI.


Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: