Hacker News new | past | comments | ask | show | jobs | submit | giuscri's comments login

as someone that quit almost 1 year ago, I can tell you that yes smokers could stare at their screen instead of talking to you when going out to have a puff


Help me seeing that way. Good references?


that’s fascinating. i don’t know if it’s the opposite to me but reading aloud can definitely help me focus.

but there’s a stigma around that: smart people just “process”.


wow, unsuited for the whole industry?


Like which one?


Can someone explain how you can interrupt with the voice this model? Where do I read more technical details about this?


How did you get exposed to Japanese media? You lived there? Or just anime? You used English subtitles or Japanese ones?


Started with friends introducing me to anime in highschool (with english subtitles), which I got hooked on, then got into the music as well, and later into vtubers (so no subtitles when watching live). I haven't ever really been into other entertainment, so for a little over a decade I've been listening to japanese on a near daily basis.

I know it's a meme for people to claim to know japanese from watching anime, which is why I don't claim to be able to speak it, but over time I did pick up enough that I don't need subtitles anymore. I'm slowly working on reading with practice books, wanikani etc, will eventually figure out some way to practice speaking too.


Where do I find info on how web search in LLMs work and how they’re trained to do that?


Web search is not a capability of a “bare” LLM, but in an LLM-based system it can be done by giving the LLM access to a “web search tool”, I.e essentially you instruct it to output a specific structured text (typically json but doesn’t have to be) indicating its “intent” to search, and your wrapper intercepts/detects this structured response, does the actual search and returns the results (e.g snippets from top k results) into the context of the LLM amd have it use these to respond to your question.

A similar thing can be done with external documents - your wrapper retrieves docs/fragments relevant to the query, puts them in the context and lets the LLM use them to answer the query. This is called Retrieval Augmented Generation (RAG).

The above is a highly simplified description. In the Langroid library (a multi-agent framework from ex-CMU/UW-Madison researchers) we have these and more. For example here’s a script that combines web search and RAG:

https://github.com/langroid/langroid/blob/main/examples/docq...


It's called retrieval augmented generation (RAG) and there's no extra training. The data (e.g. web search result) is given as input to the LLM.

If you search for "retrieval augmented generation" you'll find papers, tutorials, videos etc about it.


Can someone explain?


If we have AGI we won’t have to work anymore so why worrying about the future? In the meantime enjoy and try to partecipate in this new paradigm.

(Unless the AGI will come up capitalism is the best of the possible systems and they will rule them out of the equation)


I hope you watch more scifi movies and read more books. What people do with technology is basically never utopian or utopia adjacent.


I get a decent amount of that and dystopian scenarios are just our horror fantasies. I like them too but I like to have faith in people.


Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: